1 research outputs found

    Optimality conditions for abs-normal NLPs

    Get PDF
    Structured nonsmoothness is widely present in practical optimization problems. A particularly attractive class of nonsmooth problems, both from a theoretical and from an algorithmic perspective, are nonsmooth NLPs with equality and inequality constraints in abs-normal form, so-called abs-normal NLPs. In this thesis optimality conditions for this particular class are obtained. To this aim, first the theory for the case of unconstrained optimization problems in abs-normal form of Andreas Griewank and Andrea Walther is extended. In particular, similar necessary and sufficient conditions of first and second order are obtained that are directly based on classical Karush-Kuhn-Tucker (KKT) theory for smooth NLPs. Then, it is shown that the class of abs-normal NLPs is equivalent to the class of Mathematical Programs with Equilibrium Constraints (MPECs). Hence, the regularity assumption LIKQ introduced for the abs-normal NLP turns out to be equivalent to MPEC-LICQ. Moreover, stationarity concepts and optimality conditions under these regularity assumptions of linear independece type are equivalent up to technical assumptions. Next, well established constraint qualifications of Mangasarian Fromovitz, Abadie and Guignard type for MPECs are used to define corresponding concepts for abs-normal NLPs. Then, it is shown that kink qualifications and MPEC constraint qualifications of Mangasarian Fromovitz resp. Abadie type are equivalent. As it remains open if this holds for Guignard type kink and constraint qualifications, branch formulations for abs-normal NLPs and MPECs are introduced. Then, equivalence of Abadie’s and Guignard’s constraint qualifications for all branch problems hold. Throughout a reformulation of inequalities with absolute value slacks is considered. It preserves constraint qualifications of linear independence and Abadie type but not of Mangasarian Fromovitz type. For Guignard type it is still an open question but ACQ and GCQ are preserved passing over to branch problems. Further, M-stationarity and B-stationarity concepts for abs-normal NLPs are introduced and corresponding first order optimality con- ditions are proven using the corresponding concepts for MPECs. Moreover, a reformulation to extend the optimality conditions for abs-normal NLPs to those with additional nonsmooth objective functions is given and the preservation of regularity assumptions is considered. Using this, it is shown that the unconstrained abs-normal NLP always satisfies constraint qualifications of Abadie and thus Guignard type. Hence, in this special case every local minimizer satisfies the M-stationarity and B-stationarity concepts for abs-normal NLPs
    corecore